了解特征学习如何影响概括是现代深度学习理论的最重要目标之一。在这里,我们研究了学习表示的能力如何影响一类简单模型的概括性能:深贝叶斯线性神经网络接受了非结构化高斯数据的训练。通过将深层随机特征模型与所有训练所有层的深网进行比较,我们将提供详细的表征宽度,深度,数据密度和先验不匹配之间的相互作用。我们表明,在存在标签噪声的情况下,这两种模型都显示出样本的双重变化行为。如果有狭窄的瓶颈层,那么随机特征模型还可以显示模型的双重变化,而深网不显示这些分歧。随机特征模型可以具有特定的宽度,这些宽度对于在给定的数据密度下是最佳的概括,同时使神经网络尽可能宽或狭窄始终是最佳的。此外,我们表明,对内核限制学习曲线的前阶校正无法区分所有培训所有层的随机特征模型和深层网络。综上所述,我们的发现开始阐明建筑细节如何影响这种简单的深层回归模型类别的概括性能。
translated by 谷歌翻译
在这个简短的注意事项中,我们在宽两层沿着两层沿着双层神经网络中的存储容量问题和宽神经网络核限制的快速生长体内的储存容量问题之间的联系。具体地,我们观察到统计力学文献中研究的“有效订单参数”与无限宽度神经网络高斯工艺内核相当等同。这封对应关系连接宽两层神经网络的表达性和培训性。
translated by 谷歌翻译
深度贝叶斯神经网络中的推断仅在无限宽度的限制中完全理解,其中通过增加深度洗掉的后柔性和后部预测到浅高斯工艺。在这里,我们将有限的深度线性贝叶斯神经网络解释为跨输出通道的高斯过程预测器的数据依赖性比例混合物。我们利用这一观察来研究这些网络中的代表学习,允许我们在统一框架内的先前研究中连接限制结果。总的来说,这些结果推进了对深度影响推论的分析理解,这是一类简单的贝叶斯神经网络的推论。
translated by 谷歌翻译
最近的作品表明,有限的贝叶斯神经网络有时可能会越优于其无限堂兄弟,因为有限网络可以灵活地调整其内部表示。然而,我们对有限网络的学习隐藏层表示如何与无限网络的固定表示不同的理论理解仍然不完整。研究了对网络的扰动有限宽度校正,但已经研究过的网络,但学习特征的渐近学尚未完全表征。在这里,我们认为具有线性读数和高斯可能性的任何贝叶斯网络的平均特征内核的领先有限宽度校正具有很大程度上的普遍形式。我们明确地说明了三个易行网络架构:深线性完全连接和卷积网络,以及具有单个非线性隐藏层的网络。我们的结果开始阐明任务相关的学习信号如何塑造宽阔的贝叶斯神经网络的隐藏层表示。
translated by 谷歌翻译
Due to the high activation sparsity and use of accumulates (AC) instead of expensive multiply-and-accumulates (MAC), neuromorphic spiking neural networks (SNNs) have emerged as a promising low-power alternative to traditional DNNs for several computer vision (CV) applications. However, most existing SNNs require multiple time steps for acceptable inference accuracy, hindering real-time deployment and increasing spiking activity and, consequently, energy consumption. Recent works proposed direct encoding that directly feeds the analog pixel values in the first layer of the SNN in order to significantly reduce the number of time steps. Although the overhead for the first layer MACs with direct encoding is negligible for deep SNNs and the CV processing is efficient using SNNs, the data transfer between the image sensors and the downstream processing costs significant bandwidth and may dominate the total energy. To mitigate this concern, we propose an in-sensor computing hardware-software co-design framework for SNNs targeting image recognition tasks. Our approach reduces the bandwidth between sensing and processing by 12-96x and the resulting total energy by 2.32x compared to traditional CV processing, with a 3.8% reduction in accuracy on ImageNet.
translated by 谷歌翻译
The evaluation of object detection models is usually performed by optimizing a single metric, e.g. mAP, on a fixed set of datasets, e.g. Microsoft COCO and Pascal VOC. Due to image retrieval and annotation costs, these datasets consist largely of images found on the web and do not represent many real-life domains that are being modelled in practice, e.g. satellite, microscopic and gaming, making it difficult to assert the degree of generalization learned by the model. We introduce the Roboflow-100 (RF100) consisting of 100 datasets, 7 imagery domains, 224,714 images, and 805 class labels with over 11,170 labelling hours. We derived RF100 from over 90,000 public datasets, 60 million public images that are actively being assembled and labelled by computer vision practitioners in the open on the web application Roboflow Universe. By releasing RF100, we aim to provide a semantically diverse, multi-domain benchmark of datasets to help researchers test their model's generalizability with real-life data. RF100 download and benchmark replication are available on GitHub.
translated by 谷歌翻译
Efficient data transfers over high-speed, long-distance shared networks require proper utilization of available network bandwidth. Using parallel TCP streams enables an application to utilize network parallelism and can improve transfer throughput; however, finding the optimum number of parallel TCP streams is challenging due to nondeterministic background traffic sharing the same network. Additionally, the non-stationary, multi-objectiveness, and partially-observable nature of network signals in the host systems add extra complexity in finding the current network condition. In this work, we present a novel approach to finding the optimum number of parallel TCP streams using deep reinforcement learning (RL). We devise a learning-based algorithm capable of generalizing different network conditions and utilizing the available network bandwidth intelligently. Contrary to rule-based heuristics that do not generalize well in unknown network scenarios, our RL-based solution can dynamically discover and adapt the parallel TCP stream numbers to maximize the network bandwidth utilization without congesting the network and ensure fairness among contending transfers. We extensively evaluated our RL-based algorithm's performance, comparing it with several state-of-the-art online optimization algorithms. The results show that our RL-based algorithm can find near-optimal solutions 40% faster while achieving up to 15% higher throughput. We also show that, unlike a greedy algorithm, our devised RL-based algorithm can avoid network congestion and fairly share the available network resources among contending transfers.
translated by 谷歌翻译
With most technical fields, there exists a delay between fundamental academic research and practical industrial uptake. Whilst some sciences have robust and well-established processes for commercialisation, such as the pharmaceutical practice of regimented drug trials, other fields face transitory periods in which fundamental academic advancements diffuse gradually into the space of commerce and industry. For the still relatively young field of Automated/Autonomous Machine Learning (AutoML/AutonoML), that transitory period is under way, spurred on by a burgeoning interest from broader society. Yet, to date, little research has been undertaken to assess the current state of this dissemination and its uptake. Thus, this review makes two primary contributions to knowledge around this topic. Firstly, it provides the most up-to-date and comprehensive survey of existing AutoML tools, both open-source and commercial. Secondly, it motivates and outlines a framework for assessing whether an AutoML solution designed for real-world application is 'performant'; this framework extends beyond the limitations of typical academic criteria, considering a variety of stakeholder needs and the human-computer interactions required to service them. Thus, additionally supported by an extensive assessment and comparison of academic and commercial case-studies, this review evaluates mainstream engagement with AutoML in the early 2020s, identifying obstacles and opportunities for accelerating future uptake.
translated by 谷歌翻译
卷积神经网络(CNN)是计算机视觉(CV)中最受欢迎的人工神经网络(ANN)的模型之一。研究人员开发了各种基于CNN的结构,以解决图像分类,对象检测和图像相似性测量等问题。尽管CNN在大多数情况下显示出其价值,但它们仍然有缺点:当数据集中没有足够的样本时,它们很容易过度。大多数医疗图像数据集是此类数据集的示例。此外,许多数据集还包含设计的功能和图像,但是CNN只能直接处理图像。这是一个错过的机会来利用其他信息。因此,我们提出了一种基于CNN的模型的新结构:Compnet,一个复合卷积神经网络。这是一个专门设计的神经网络,可以接受图像和设计功能的组合作为输入,以利用所有可用信息。这种结构的新颖性是,它使用从图像到重量设计的功能学习的功能,以便从图像和设计功能中获取所有信息。随着该结构在分类任务上的使用,结果表明我们的方法有能力显着减少过度拟合。此外,我们还发现了其他研究人员提出的几种类似的方法,可以结合图像和设计功能。为了进行比较,我们首先在LIDC上应用了这些类似的方法,并将结果与​​Compnet结果进行了比较,然后我们将COMPNET应用于数据集中,这些方法最初在其作品中最初使用,并将结果与​​他们在论文中提出的结果进行了比较。 。所有这些比较结果表明,我们的模型在LIDC数据集或其提议的数据集上的分类任务上优于这些类似的方法。
translated by 谷歌翻译
几种慢性肺疾病,例如特发性肺纤维化(IPF)的特征是气道异常扩张。计算机断层扫描(CT)上气道特征的定量可以帮助表征疾病进展。已经开发了基于物理的气道测量算法,但由于在临床实践中看到的气道形态多样性,因此取得了有限的成功。由于获得精确的气道注释的高成本,监督学习方法也不可行。我们建议使用感知损失通过样式转移进行综合气道,以训练我们的模型气道转移网络(ATN)。我们使用a)定性评估将ATN模型与最先进的GAN网络(SIMGAN)进行比较; b)评估基于ATN和SIMGAN的CT气道指标预测113例IPF患者死亡率的能力。与Simgan相比,ATN被证明更快,更容易训练。还发现基于ATN的气道测量值始终比IPF CTS上的SIMGAN衍生气道指标更强大。通过转化网络使用感知损失来完善合成数据的转化网络是基于GAN的方法的现实替代方法,用于用于特发性肺纤维化的临床CT分析。我们的源代码可以在https://github.com/ashkanpakzad/atn上找到,该源代码与Airquant的现有开放源气道分析框架兼容。
translated by 谷歌翻译